翻訳と辞書
Words near each other
・ "O" Is for Outlaw
・ "O"-Jung.Ban.Hap.
・ "Ode-to-Napoleon" hexachord
・ "Oh Yeah!" Live
・ "Our Contemporary" regional art exhibition (Leningrad, 1975)
・ "P" Is for Peril
・ "Pimpernel" Smith
・ "Polish death camp" controversy
・ "Pro knigi" ("About books")
・ "Prosopa" Greek Television Awards
・ "Pussy Cats" Starring the Walkmen
・ "Q" Is for Quarry
・ "R" Is for Ricochet
・ "R" The King (2016 film)
・ "Rags" Ragland
・ ! (album)
・ ! (disambiguation)
・ !!
・ !!!
・ !!! (album)
・ !!Destroy-Oh-Boy!!
・ !Action Pact!
・ !Arriba! La Pachanga
・ !Hero
・ !Hero (album)
・ !Kung language
・ !Oka Tokat
・ !PAUS3
・ !T.O.O.H.!
・ !Women Art Revolution


Dictionary Lists
翻訳と辞書 辞書検索 [ 開発暫定版 ]
スポンサード リンク

fuzzy clustering : ウィキペディア英語版
fuzzy clustering

Data clustering is the process of dividing data elements into classes or clusters so that items in the same class are as similar as possible, and items in different classes are as dissimilar as possible. Depending on the nature of the data and the purpose for which clustering is being used, different measures of similarity may be used to place items into classes, where the similarity measure controls how the clusters are formed. Some examples of measures that can be used as in clustering include distance, connectivity, and intensity.
In hard clustering, data is divided into distinct clusters, where each data element belongs to exactly one cluster. In fuzzy clustering (also referred to as soft clustering), data elements can belong to more than one cluster, and associated with each element is a set of membership levels. These indicate the strength of the association between that data element and a particular cluster. Fuzzy clustering is a process of assigning these membership levels, and then using them to assign data elements to one or more clusters.
One of the most widely used fuzzy clustering algorithms is the Fuzzy C-Means (FCM) Algorithm
(Bezdek 1981). The FCM algorithm attempts to partition a finite collection of n elements
X = \_n \} into a collection of c fuzzy clusters with respect to some given criterion.
Given a finite set of data, the algorithm returns a list of c cluster centres C = \_c \} and a partition matrix W = w_ \in(1 ),\; i = 1, . . . , n,\; j = 1, . . . , c, where each element w_ tells
the degree to which element \mathbf_i belongs to cluster \mathbf_j. Like the K-means clustering, the FCM
aims to minimize an objective function:
:\underset ^ \sum_^ w_^m \left\|\mathbf_i - \mathbf_j \right\|^2,
where:
:w^m_ = \frac \left(\frac_j \right\|}_k \right\|}\right)^}}.
This differs from the ''k''-means objective function by the addition of the membership values w_ and the fuzzifier m \in R, with m \ge 1. The fuzzifier m determines the level of cluster fuzziness. A large m results in smaller memberships w_ and hence, fuzzier clusters. In the limit m=1, the memberships w_ converge to 0 or 1, which implies a crisp partitioning. In the absence of experimentation or domain knowledge, m is commonly set to 2.
== Fuzzy c-means clustering ==

In fuzzy clustering, every point has a degree of belonging to clusters, as in fuzzy logic, rather than belonging completely to just one cluster. Thus, points on the edge of a cluster, may be ''in the cluster'' to a lesser degree than points in the center of cluster. An overview and comparison of different fuzzy clustering algorithms is available.〔Nock, R. and Nielsen, F. (2006) ( "On Weighting Clustering" ), IEEE Trans. on Pattern Analysis and Machine Intelligence, 28 (8), 1–13〕
Any point ''x'' has a set of coefficients giving the degree of being in the ''k''th cluster ''w''''k''(''x''). With fuzzy ''c''-means, the centroid of a cluster is the mean of all points, weighted by their degree of belonging to the cluster:
:c_k = }.
The degree of belonging, ''w''''k''(''x''), is related inversely to the distance from ''x'' to the cluster center as calculated on the previous pass. It also depends on a parameter ''m'' that controls how much weight is given to the closest center. The fuzzy ''c''-means algorithm is very similar to the ''k''-means algorithm:
* Choose a number of clusters.
* Assign randomly to each point coefficients for being in the clusters.
* Repeat until the algorithm has converged (that is, the coefficients' change between two iterations is no more than \varepsilon, the given sensitivity threshold) :
*
* Compute the centroid for each cluster, using the formula above.
*
* For each point, compute its coefficients of being in the clusters, using the formula above.
The algorithm minimizes intra-cluster variance as well, but has the same problems as ''k''-means; the minimum is a local minimum, and the results depend on the initial choice of weights.
Using a mixture of Gaussians along with the expectation-maximization algorithm is a more statistically formalized method which includes some of these ideas: partial membership in classes.
Another algorithm closely related to Fuzzy C-Means is Soft K-means.
Fuzzy c-means has been a very important tool for image processing in clustering objects in an image. In the 70's, mathematicians introduced the spatial term into the FCM algorithm to improve the accuracy of clustering under noise.〔.〕

抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)
ウィキペディアで「fuzzy clustering」の詳細全文を読む



スポンサード リンク
翻訳と辞書 : 翻訳のためのインターネットリソース

Copyright(C) kotoba.ne.jp 1997-2016. All Rights Reserved.